SmartNotes: Implicit Labeling of Meeting Data through User Noteâ•fiTaking and Browsing
نویسندگان
چکیده
We have built SmartNotes, a system with two goals: first to help users take multi– media notes during meetings and browse those notes afterwards, and second to acquire labeled data from users’ interactions with the system with which to later improve the performance of components that perform automatic meeting understanding. SmartNotes consists of a stand–alone laptop based note taking system, and a web based note retrieval system. We shall demonstrate both systems, and will also show the kinds of labeled data that get created during typical meetings and browsing sessions. 1 Goals of the SmartNotes System Most institutions hold a large number of meetings every day. Of these meetings, many are important, where we can define a meeting as important if at a future date someone in the institution needs to recall details discussed at the meeting. In a previous survey (Banerjee et al., 2005) of busy professors at Carnegie Mellon University we showed that on average meeting participants needed to recall details of past meetings about twice a month. Performing such retrieval is not an easy task. It is time consuming (in our study participants took on average between 15 minutes to an hour to recall the information), and worse, the results of the retrieval are unsatisfactory (satisfaction rates varied between 3.4 to 4 on a scale of 0 to 5, with 5 being completely satisfactory). Despite the prevalance of important meetings and the costs of retrieving information from them afterwards, there is, however, a relative paucity of techonology to help meeting participants record and retrieve information discussed at meetings. Our first goal is to build a system that helps meeting participants easily create multi–media notes (text notes augmented with audio/video recordings of the meeting), and access them easily after the meeting. Such a system can make meeting note–taking a collaborative process between the different meeting participants and the system, with the system performing the “easier” note–taking tasks (such as automatically recording the clearly stated action items) and allowing the participants to perform the harder tasks. To perform such note taking, the system needs to form an understanding of the meeting. This task currently includes 3 components: automatic topic detection, action item detection and participant role detection. We would further like these components to improve their performance over time by learning through observation of the user’s note–taking and note–retrieval activity. Thus our second goal is to set up the system in such a way that the user’s interactions with the system result in labeled meeting data that can then be used to improve the meeting understanding components. Towards these two goals, we have built SmartNotes which helps users easily record and retrieve notes, and records those user interactions to form labeled meeting data that can later be used to automatically improve the meeting understanding components. In the next section we describe the three meeting understanding components in more detail. Next we describe SmartNotes itself, and show how it is currently helping users take and retrieve notes. We also demonstrate how it is acquiring labeled data to aid each of the meeting understanding componenets. Finally we end with a discussion of what functionality we plan to demonstrate at the conference. 2 Automatic Meeting Understanding We are currently performing research into three kinds of automatic meeting understanding tasks, as follows. Topic detection and segmentation: We are attempting to automatically discover the topics being discussed at meetings. This task is composed of two sub-tasks: discovering the points in a meeting when the topic changes, and then associating a descriptive label to the segment between two topic shifts. Our current strategy for topic shift detection is to perform an edge detection (like in (Hearst, 1997)) using such features as speech activity (who spoke when and for how long), the words that each person spoke, etc. For labeling, we are currently simply associating the agenda item names recorded in the notes with the segments they are most relevant to, as decided by a tf.idf matching technique. Topic detection is particularly useful during meeting information retrieval; (Banerjee et al., 2005) showed that when users wish to retrieve information from past meetings, they are typically interested in a specific discussion topic, as opposed to an entire meeting. Action item detection: An obvious application of meeting understanding is the automatic discovery and recording of action items as they are discussed during a meeting. Arguably one of the most important outcomes of a meeting are the action items decided upon, and automatically recording them could be a huge benefit especially to those participants that are likely to not note them down and consequently forget about them later on. Meeting participant role detection: Each meeting participant plays a variety of roles in an institution. These roles could be based on their function in the institution (managers, assitants, professors, students, etc), or based on their expertise (speech recognition experts, facilities experts, etc). Our current strategy for role detection is to train detectors on hand labeled data. Our next step is to perform disFigure 1: Screen shot of the SmartNotes note–taking client covery of new roles through clustering techniques. Detecting such roles has several benefits. First, it allows us to build prior expectations of a meeting between a group of participants. For example, if we know person A is a speech recognition expert and person B a speech synthesis expert, a reasonable expectation is that when they meet they are likely to talk about technologies related speech processing. Consequently, we can use this expectation to aid the action item detection and the topic detection in that meeting. 3 SmartNotes: System Description We have built SmartNotes to help users take multi– media notes during meetings, and retrieve them later on. SmartNotes consists of two major components: The note taking application which meeting participants use to take notes during the meeting, and the note retrieval application which users use to retrieve notes at a later point. 3.1 SmartNotes Note Taking Application The note taking application is a stand–alone system, that runs on each meeting participant’s laptop, and allows him to take notes during the meeting. In addition to recording the text notes, it also records the participant’s speech, and video, if a video camera is connected to the laptop. This system is an extension of the Carnegie Mellon Meeting Recorder (Banerjee et al., 2004). Figure 1 shows a screen–shot of this application. It is a server–client application, and each participant logs into a central server at the beginning of each meeting. Thus, the system knows the precise identity of each note taker as well as each speaker in the meeting. This allows us to avoid the onerous problem of automatically detecting who is speaking at any time during the meeting. Further, after logging on, each client automatically synchronizes itself with a central NTP time server. Thus the time stamps that each client associates with its recordings are all synchronized, to facilitate merging and play back of audio/video during browsing (described in the next sub–section). Once logged in, each participant’s note taking area is split into two sections: a shared note taking area, and a private note taking area. Notes written in the shared area are viewable by all meeting participants. This allows meeting participants to share the task of taking notes during a meeting: As long as one participant has recorded an important point during a meeting, the other participants do not need to, thus making the note taking task easier for the group as a whole. Private notes that a participant does not wish to share with all participants can be taken in the private note taking area. The interface has a mechanism to allow meeting participants to insert an agenda into the shared area. Once inserted, the shared area is split into as many boxes as there are agenda items. Participants can then take notes during the discussion of an agenda item in the corresponding agenda item box. This is useful to the participants because it organizes the notes as they are being taken, and, additionally, the notes can later be retrieved agenda item by agenda item. Thus, the user can access all notes he has taken in different meetings regarding “buying a printer”, without having to see the notes taken for the other agenda items in each such meeting. In addition to being useful to the user, this act of inserting an agenda and then taking notes within the relevant agenda item box results in generating (unbeknownst to the participant) labeled data for the topic detection component. Specifically, if we define each agenda item as being a separate “topic”, and make the assumption that notes are taken approximately concurrent with the discussion of the contents of the notes, then we can conclude that there is a shift in the topic of discussion at some point between the time stamp on the last note in an agenda item box, and the time stamp on the first note of the next agenda item box. This information can then be used to improve the performance of the topic shift detector. The accuracy of the topic shift data thus acquired depends on the length of time between the two time points. Since this length is easy to calculate automatically, this information can be factored into the topic detector trainer. The interface also allows participants to enter action items through a dedicated action item form. Again the advantage of such a form to the participants is that the action items (and thus the notes) are better organized: After the meeting, they can perform retrieval on specific fields of the action items. For example, they can ask to retrieve all the action items assigned to a particular participant, or that are due a particular day, etc. In addition to being beneficial to the participant, the action item form filling action results in generating (unbeknownst to the participant) labeled data for the action item detector. Specifically, if we make the assumption that an action item form filling action is preceded by a discussion of the action item, then the system can couple the contents of the form with all the speech within a window of time before the form filling action, and use this pair as a data point to retrain its actiom item detector. 3.2 SmartNotes Note Retrieval Website As notes and audio/video are recorded on each individual participant’s laptop, they also get transferred over the internet to a central meeting server. This transfer occurs in the background without any intervention from the user, utilizes only the left–over bandwidth beyond the user’s current bandwidth usage, and is robust to system shut–downs, crashes, etc. This process is described in more detail in (Banerjee et al., 2004). Once the meeting is over and all the data has been transferred to the central server, meeting participants can use the SmartNotes multi–media notes retrieval system to view the notes and access the recorded audio/video. This is a web–based application that uses the same login process as the stand–along note taking system. Users can view a list of meetings they have recorded using the SmartNotes application in the past, and then for each meeting, they can Figure 2: Screen shot of the SmartNotes website view the shared notes taken at the meeting. Figure 2 shows a screen shot of such a notes browsing session. Additionally, participants can view their own private notes taken during the meeting. In addition to viewing the notes, they can also access all recorded audio/video, indexed by the notes. That is, they can access the audio/video recorded around the time that the note was entered. Further they can specify how many minutes before and after the note they wish to access. Since the server has the audio from each meeting participant’s audio channel, the viewer of the notes can choose to listen to any one person’s channel, or a combination of the audio channels. The merging of channels is done in real time and is achievable because their time stamps have been synchronized during recording. Finally we plan to implement a simple key– word based search on the notes recorded in all the recorded meetings (or in one specific meeting) by the time of the demonstration in June. This search will return notes that match the search using a standard tf.idf approach. The user will also be provided the option of rating the quality of the search retrieval on a one bit satisfied/not–satisfied scale. If the user chooses to provide this rating, it can be used as a feedback to improve the search. Additionally, which parts of the meeting the user chooses to access the audio/video from can be used to form a model of the parts of the meetings most relevant to the user. This information can help the system tailor its retrieval to individual preferences. 4 The Demonstration If accepted, we shall demonstrate both the SmartNotes note taking client as well as the SmartNotes note–retrieval website. Specifically we will perform 2 minute long mock meetings between 2 or 3 demonstrators or audience participants. We will show how the notes can be taken, how agendas can be created and action items noted. We will then show how that 2 minute meeting immediately appears on the SmartNotes note retrieval website, how the notes can be browsed, and the audio accessed. We shall show the automatically labeled data that was created both during the mock meeting, as well as during the browsing session. Finally, if time permits, we shall show results on how much we can improve the meeting understanding components’ capabilities through labeled meeting data automatically acquired through participants’ use of SmartNotes at CMU and other institutions that are currently using the system. To perform this demonstration we shall require power supplies to support at least 3 laptops (for the 3 meeting participants during the mock meeting demonstration), as well as internet access (preferably wired) for each of the laptops. Additionally we need a projector and a projection screen to display the interfaces, and two easels to display posters that describe the system and the research results obtained based on the system.
منابع مشابه
SmartNotes: Implicit Labeling of Meeting Data through User Note-Taking and Browsing
We have implemented SmartNotes, a system that automatically acquires labeled meeting data as users take notes during meetings and browse the notes afterwards. Such data can enable meeting understanding components such as topic and action item detectors to automatically improve their performance over a sequence of meetings. The SmartNotes system consists of a laptop based note taking application...
متن کاملUser Centred Evaluation of A Recommendation Based Image Browsing System
In this paper, we introduce a novel approach to recommend images by mining user interactions based on implicit feedback of user browsing. The underlying hypothesis is that the interaction implicitly indicates the interests of the users for meeting practical image retrieval tasks. The algorithm mines interaction data and also low-level content of the clicked images to choose diverse images by cl...
متن کاملA Framework for Structuring Multimedia Archives and for Browsing Efficiently through Multimodal Links
This thesis proposes a method for indexing and browsing archives of multimedia documents, and in particular meeting recordings, using printable documents and links. Existing systems for indexing and browsing multimedia data have four main limits. First, the indexing requires high-level abstractions extracted from multimedia documents, which is still an unsolved problem for rich media such as im...
متن کاملAn Ego-Centric and Tangible Approach to Meeting Indexing and Browsing
This article presents an ego-centric approach for indexing and browsing meetings. The method considers two concepts: meetings’ data alignment with personal information to enable ego-centric browsing and live intentional annotation of meetings through tangible actions to enable ego-centric indexing. The article first motivates and introduces these concepts and further presents brief states-of-th...
متن کاملA Novel Trust Computation Method Based on User Ratings to Improve the Recommendation
Today, the trust has turned into one of the most beneficial solutions to improve recommender systems, especially in the collaborative filtering method. However, trust statements suffer from a number of shortcomings, including the trust statements sparsity, users' inability to express explicit trust for other users in most of the existing applications, etc. Thus to overcome these problems, this ...
متن کامل